Eager Stack Cache Memory Transfers
نویسندگان
چکیده
The growing complexity of modern computer architectures increasingly complicates the prediction of the run-time behavior of software. For real-time systems, where a safe estimation of the program’s worst-case execution time is needed, time-predictable computer architectures promise to resolve this problem. The stack cache, for instance, allows the compiler to efficiently cache a program’s stack, while static analysis of its behavior remains easy. This work introduces an optimization of the stack cache that allows to anticipate memory transfers that might be initiated by future stack cache control instructions. These eager memory transfers thus allow to reduce the average-case latency of those control instructions, very similar to “prefetching” techniques known from conventional caches. However, the mechanism proposed here is guaranteed to have no impact on the worstcase execution time estimates computed by static analysis. Measurements on a dual-core platform using the Patmos processor and timedivision-multiplexing-based memory arbitration, show that our technique can eliminate up to 62% (7%) of the memory transfers from (respectively to) the stack cache on average over all programs of the MiBench benchmark suite. 1998 ACM Subject Classification C.3 [Special-Purpose and Application-Based Systems] RealTime and Embedded Systems
منابع مشابه
Improving Bandwidth Utilization using Eager Writeback
Cache memories have been incorporated into almost all modern, general-purpose microprocessors. To maintain data consistency between cache structures and the rest of the memory systems, most of these caches employ either a writeback or a write-through strategy to deal with store operations. Writethrough caches propagate data to more distant memory levels at the time each store occurs, producing ...
متن کاملReducing DRAM Row Activations with Eager Writeback
Reducing DRAM Row Activations with Eager Writeback
متن کاملSleepy Stack: a New Approach to Low Power VLSI Logic and Memory
Outline zIntroduction zRelated Work zSleepy stack zSleepy stack logic circuits zSleepy stack SRAM zLow-power pipelined cache (LPPC) zSleepy stack pipelined SRAM zConclusion
متن کاملLV*: A low complexity lazy versioning HTM infrastructure
Transactional memory (TM) promises to unlock parallelism in software in a safer and easier way than lock-based approaches but the path to deployment is unclear for several reasons. First of all, since TM has not been deployed in any machine yet, experience of using it is limited. While software transactional memory implementations exist, they are too slow to provide useful experience. Existing ...
متن کاملEager Data Transfer Mechanism for Reducing Communication Latency in User-Level Network Protocols
Clusters have become a popular alternative for building high-performance parallel computing systems. Today's high-performance system area network (SAN) protocols such as VIA and IBA significantly reduce user-to-user communication latency by implementing protocol stacks outside of operating system kernel. However, emerging parallel applications require a significant improvement in communication ...
متن کامل